Warped Convolutions: Efficient Invariance to Spatial Transformations
نویسندگان
چکیده
Convolutional Neural Networks (CNNs) are extremely efficient, since they exploit the inherent translation-invariance of natural images. However, translation is just one of a myriad of useful spatial transformations. Can the same efficiency be attained when considering other spatial invariances? Such generalized convolutions have been considered in the past, but at a high computational cost. We present a construction that is simple and exact, yet has the same computational complexity that standard convolutions enjoy. It consists of a constant image warp followed by a simple convolution, which are standard blocks in deep learning toolboxes. With a carefully crafted warp, the resulting architecture can be made invariant to one of a wide range of spatial transformations. We show encouraging results in realistic scenarios, including the estimation of vehicle poses in the Google Earth dataset (rotation and scale), and face poses in Annotated Facial Landmarks in the Wild (3D rotations under perspective).
منابع مشابه
HexaConv
The effectiveness of convolutional neural networks stems in large part from their ability to exploit the translation invariance that is inherent in many learning problems. Recently, it was shown that CNNs can exploit other sources of invariance, such as rotation invariance, by using group convolutions instead of planar convolutions. However, for reasons of performance and ease of implementation...
متن کاملDiscriminative template learning in group-convolutional networks for invariant speech representations
In the framework of a theory for invariant sensory signal representations, a signature which is invariant and selective for speech sounds can be obtained through projections in template signals and pooling over their transformations under a group. For locally compact groups, e.g., translations, the theory explains the resilience of convolutional neural networks with filter weight sharing and ma...
متن کاملWarped Convolutions: A Novel Tool in the Construction of Quantum Field Theories
Recently, Grosse and Lechner introduced a novel deformation procedure for non–interacting quantum field theories, giving rise to interesting examples of wedge–localized quantum fields with a non–trivial scattering matrix. In the present article we outline an extension of this procedure to the general framework of quantum field theory by introducing the concept of warped convolutions: given a th...
متن کاملAlgebraic constructive quantum field theory: Integrable models and deformation techniques
Several related operator-algebraic constructions for quantum field theory models on Minkowski spacetime are reviewed. The common theme of these constructions is that of a Borchers triple, capturing the structure of observables localized in a Rindler wedge. After reviewing the abstract setting, we discuss in this framework i) the construction of free field theories from standard pairs, ii) the i...
متن کاملLearning Invariance with Compact Transforms
The problem of building machine learning models that admit efficient representations and also capture an appropriate inductive bias for the domain has recently attracted significant interest. Existing work for compressing deep learning pipelines has explored classes of structured matrices that exhibit forms of shift-invariance akin to convolutions. We leverage the displacement rank framework to...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2017